Local Multidimensional Scaling for Nonlinear Dimension Reduction, Graph Layout and Proximity Analysis
نویسندگان
چکیده
In recent years there has been a resurgence of interest in nonlinear dimension reduction methods. Among new proposals are so-called “Local Linear Embedding” (LLE) and “Isomap”. Both use local neighborhood information to construct a global lowdimensional embedding of a hypothetical manifold near which the data fall. In this paper we introduce a family of new nonlinear dimension reduction methods called “Local Multidimensional Scaling” or LMDS. Like LLE and Isomap, LMDS only uses local information from user-chosen neighborhoods, but it differs from them in that it uses ideas from the area of “graph layout”. A common paradigm in graph layout is to achieve desirable drawings of graphs by minimizing energy functions that balance attractive forces between near points and repulsive forces between non-near points against each other. We approach the force paradigm by proposing a parametrized family of stress or energy functions inspired by Box-Cox power transformations. This family provides users with considerable flexibility for achieving desirable embeddings, and it comprises most energy functions proposed in the past. Facing an embarrassment of riches of energy functions, however, one needs a method for selecting viable energy functions. We solve this problem by proposing a metacriterion that measures how well the sets of K-nearest neighbors agree between the original high-dimensional space and the low-dimensional embedding space. This metacriterion has intuitive appeal, and it performs well in creating faithful embeddings. The meta-criterion not only solves the problem of choice of energy function but also provides a measure for comparing embeddings generated by arbitrary dimension reduction methods, including principal components, LLE and Isomap. Lastly, the meta-criterion
منابع مشابه
Local Multidimensional Scaling for Nonlinear Dimension Reduction, Graph Drawing and Proximity Analysis
In the past decade there has been a resurgence of interest in nonlinear dimension reduction. Among new proposals are “Local Linear Embedding” (LLE, Roweis and Saul 2000), “Isomap” (Tenenbaum et al. 2000) and Kernel PCA (KPCA, Schölkopf, Smola and Müller 1998), which all construct global lowdimensional embeddings from local affine or metric information. We introduce a competing method called “Lo...
متن کاملXGvis: Interactive Data Visualization with Multidimensional Scaling
We discuss interactive techniques for multidimensional scaling (MDS) and a system, named \XGvis", that implements these techniques. MDS is a method for visualizing proximity data, that is, data where objects are characterized by dissimilarity values for all pairs of objects. MDS constructs maps of these objects in IR k by interpreting the dissimilarities as distances. MDS in its conventional ba...
متن کاملEnergy/Stress Functions for Dimension Reduction and Graph Drawing: Power Laws and Their Clustering Properties∗
We introduce a parametrized family of energy/stress functions useful for proximity analysis, nonlinear dimension reduction, and graph drawing. The functions are inspired by the physics intuitions of attractive and repulsive forces common in graph drawing. Their minimization generates low-dimensional configurations (embeddings, graph drawings) whose interpoint distances match input distances as ...
متن کاملStress functions for nonlinear dimension reduction, proximity analysis, and graph drawing
Multidimensional scaling (MDS) is the art of reconstructing pointsets (embeddings) from pairwise distance data, and as such it is at the basis of several approaches to nonlinear dimension reduction and manifold learning. At present, MDS lacks a unifying methodology as it consists of a discrete collection of proposals that differ in their optimization criteria, called “stress functions”. To corr...
متن کاملIntegrating Spatial Proximity with Manifold Learning
Dimension reduction is a useful preprocessing step for many types hyperspectral image analysis, including visualization, regression, clustering and classification. By dimension reduction, high dimensional data are mapped into a lower dimensional space while the important features of the original data are preserved according to a given criterion. Although linear dimension reduction methods such ...
متن کامل